Dropping Convexity for Faster Semi-definite Optimization
نویسندگان
چکیده
We study the minimization of a convex function f(X) over the set of n × n positive semi-definite matrices, but when the problem is recast as minU g(U) := f(UU >), with U ∈ Rn×r and r ≤ n. We study the performance of gradient descent on g—which we refer to as Factored Gradient Descent (Fgd)—under standard assumptions on the original function f . We provide a rule for selecting the step size and, with this choice, show that the local convergence rate of Fgd mirrors that of standard gradient descent on the original f : i.e., after k steps, the error is O(1/k) for smooth f , and exponentially small in k when f is (restricted) strongly convex. In addition, we provide a procedure to initialize Fgd for (restricted) strongly convex objectives and when one only has access to f via a first-order oracle; for several problem instances, such proper initialization leads to global convergence guarantees. Fgd and similar procedures are widely used in practice for problems that can be posed as matrix factorization. To the best of our knowledge, this is the first paper to provide precise convergence rate guarantees for general convex functions under standard convex assumptions.
منابع مشابه
A Note on the Convexity of $\log \det ( I + KX^{-1} )$ and its Constrained Optimization Representation
This note provides another proof for the convexity (strict convexity) of log det(I + KX) over the positive definite cone for any given positive semidefinite matrix K 0 (positive definite matrix K ≻ 0) and the strictly convexity of log det(K + X) over the positive definite cone for any given K 0. Equivalent optimization representation with linear matrix inequalities (LMIs) for the functions log ...
متن کاملPositive Semidefiniteness and Positive Definiteness of a Linear Parametric Interval Matrix
We consider a symmetric matrix, the entries of which depend linearly on some parameters. The domains of the parameters are compact real intervals. We investigate the problem of checking whether for each (or some) setting of the parameters, the matrix is positive definite (or positive semidefinite). We state a characterization in the form of equivalent conditions, and also propose some computati...
متن کاملA New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization
We present a new full Nesterov and Todd step infeasible interior-point algorithm for semi-definite optimization. The algorithm decreases the duality gap and the feasibility residuals at the same rate. In the algorithm, we construct strictly feasible iterates for a sequence of perturbations of the given problem and its dual problem. Every main iteration of the algorithm consists of a feasibili...
متن کاملConvexity conditions of Kantorovich function and related semi-infinite linear matrix inequalities
The Kantorovich function (xT Ax)(xT A−1x), where A is a positive definite matrix, is not convex in general. From a matrix or convex analysis point of view, it is interesting to address the question: When is this function convex? In this paper, we prove that the 2dimensional Kantorovich function is convex if and only if the condition number of its matrix is less than or equal to 3 + 2 √ 2. Thus ...
متن کاملConvexity of some spectral functions on Hermitian matrices
We prove in this note the convexity of the functions u ◦ λ and more generally u ◦ λB on the space of Hermitian matrices, for B a fixed positive definite hermitian matrix, when u : R → R ∪ {+∞} is a symmetric convex function which is lower semi-continuous on R, and finite in at least one point of R. This is performed by using some optimisation techniques and a generalized Ky Fan inequality. To c...
متن کامل